Goto

Collaborating Authors

 Devon



Royal Navy returns to wind power with trial of robotic sailboats

New Scientist

Oshen's robotic sailboats are powered by the wind and the sun The UK's Royal Navy may return to the age of sail, with a new demonstration involving a flotilla of small, wind-propelled robot boats. Made by Oshen in Plymouth, UK, the vessels, known as C-Stars, are just 1.2 metres long and weigh around 40 kilos. Solar panels power navigation, communications and sensors, while a sail provides propulsion. Deployed as a constellation, the small vessels act as a wide-area sensor network. How the US military wants to use the world's largest aircraft "The simplest way of describing C-Stars is as self-deploying, station-keeping ocean buoys," says Oshen CEO Anahita Laverack .


Can a new book crack one of neuroscience's hardest problems? Not quite

New Scientist

The ideas presented in George Lakoff and Srini Narayanan's The Neural Mind are fascinating, but the writing is far less compelling This is a book review in two parts. The first is about the ideas presented in The Neural Mind: How brains think, which are fascinating. The second is about the actual experience of reading it. The book tackles one of the biggest questions in neuroscience: how do neurons perform all the different kinds of human thought possible, from planning motor actions to composing sentences and musing about philosophy? The authors have very different perspectives.


Pumped Hydro Energy Storage Is Having a Renaissance

WIRED

As the world looks to incorporate more renewables into energy grids, centuries-old systems that can balance supply and demand are being reappraised and innovated upon.


Towards 6G Native-AI Edge Networks: A Semantic-Aware and Agentic Intelligence Paradigm

Feng, Chenyuan, Zhang, Anbang, Min, Geyong, Huang, Yongming, Quek, Tony Q. S., You, Xiaohu

arXiv.org Artificial Intelligence

The evolution toward sixth-generation wireless systems positions intelligence as a native network capability, fundamentally transforming the design of radio access networks (RANs). Within this vision, Semantic-native communication and agentic intelligence are expected to play central roles. SemCom departs from bit-level fidelity and instead emphasizes task-oriented meaning exchange, enabling compact SC and introducing new performance measures such as semantic fidelity and task success rate. Agentic intelligence endows distributed RAN entities with goal-driven autonomy, reasoning, planning, and multi-agent collaboration, increasingly supported by foundation models and knowledge graphs. In this work, we first introduce the conceptual foundations of SemCom and agentic networking, and discuss why existing AI-driven O-RAN solutions remain largely bit-centric and task-siloed. We then present a unified taxonomy that organizes recent research along three axes: i) semantic abstraction level (symbol/feature/intent/knowledge), ii) agent autonomy and coordination granularity (single-, multi-, and hierarchical-agent), and iii) RAN control placement across PHY/MAC, near-real-time RIC, and non-real-time RIC. Based on this taxonomy, we systematically introduce enabling technologies including task-oriented semantic encoders/decoders, multi-agent reinforcement learning, foundation-model-assisted RAN agents, and knowledge-graph-based reasoning for cross-layer awareness. Representative 6G use cases, such as immersive XR, vehicular V2X, and industrial digital twins, are analyzed to illustrate the semantic-agentic convergence in practice. Finally, we identify open challenges in semantic representation standardization, scalable trustworthy agent coordination, O-RAN interoperability, and energy-efficient AI deployment, and outline research directions toward operational semantic-agentic AI-RAN.


The promising potential of vision language models for the generation of textual weather forecasts

Steele, Edward C. C., Mane, Dinesh, Monti, Emilio, Orus, Luis, Chantrill-Cheyette, Rebecca, Couch, Matthew, Dale, Kirstine I., Eaton, Simon, Rangarajan, Govindarajan, Majlesi, Amir, Ramsdale, Steven, Sharpe, Michael, Smith, Craig, Smith, Jonathan, Yates, Rebecca, Ellis, Holly, Ewen, Charles

arXiv.org Artificial Intelligence

Despite the promising capability of multimodal foundation models, their application to the generation of meteorological products and services remains nascent. To accelerate aspiration and adoption, we explore the novel use of a vision language model for writing the iconic Shipping Forecast text directly from video-encoded gridded weather data. These early results demonstrate promising scalable technological opportunities for enhancing production efficiency and service innovation within the weather enterprise and beyond.


Reasoning-Aware Multimodal Fusion for Hateful Video Detection

Yang, Shuonan, Chen, Tailin, Yue, Jiangbei, Cheng, Guangliang, Jiao, Jianbo, Fu, Zeyu

arXiv.org Artificial Intelligence

Hate speech in online videos is posing an increasingly serious threat to digital platforms, especially as video content becomes increasingly multimodal and context-dependent. Existing methods often struggle to effectively fuse the complex semantic relationships between modalities and lack the ability to understand nuanced hateful content. To address these issues, we propose an innovative Reasoning-Aware Multimodal Fusion (RAMF) framework. To tackle the first challenge, we design Local-Global Context Fusion (LGCF) to capture both local salient cues and global temporal structures, and propose Semantic Cross Attention (SCA) to enable fine-grained multimodal semantic interaction. To tackle the second challenge, we introduce adversarial reasoning-a structured three-stage process where a vision-language model generates (i) objective descriptions, (ii) hate-assumed inferences, and (iii) non-hate-assumed inferences-providing complementary semantic perspectives that enrich the model's contextual understanding of nuanced hateful intent. Evaluations on two real-world hateful video datasets demonstrate that our method achieves robust generalisation performance, improving upon state-of-the-art methods by 3% and 7% in Macro-F1 and hate class recall, respectively. We will release the code after the anonymity period ends.



PriVi: Towards A General-Purpose Video Model For Primate Behavior In The Wild

Mueller, Felix B., Meier, Jan F., Lueddecke, Timo, Vogg, Richard, Freixanet, Roger L., Hassler, Valentin, Bosshard, Tiffany, Karakoc, Elif, O'Hearn, William J., Pereira, Sofia M., Sehner, Sandro, Wierucka, Kaja, Burkart, Judith, Fichtel, Claudia, Fischer, Julia, Gail, Alexander, Hobaiter, Catherine, Ostner, Julia, Samuni, Liran, Schülke, Oliver, Shahidi, Neda, Wessling, Erin G., Ecker, Alexander S.

arXiv.org Artificial Intelligence

Non-human primates are our closest living relatives, and analyzing their behavior is central to research in cognition, evolution, and conservation. Computer vision could greatly aid this research, but existing methods often rely on human-centric pretrained models and focus on single datasets, which limits generalization. W e address this limitation by shifting from a model-centric to a data-centric approach and introduce PriVi, a large-scale primate-centric video pretraining dataset. PriVi contains 424 hours of curated video, combining 174 hours from behavioral research across 11 settings with 250 hours of diverse web-sourced footage, assembled through a scalable data cura-tion pipeline. W e continue pretraining V-JEP A, a large-scale video model, on PriVi to learn primate-specific representations and evaluate it using a lightweight frozen classifier . Across four benchmark datasets - ChimpACT, PanAf500, BaboonLand, and ChimpBehave - our approach consistently outperforms prior work, including fully fine-tuned baselines, and scales favorably with fewer labels. These results demonstrate that primate-centric pretraining substantially improves data efficiency and generalization, making it a promising approach for low-label applications. Code, models, and the majority of the dataset will be made available.


BadThink: Triggered Overthinking Attacks on Chain-of-Thought Reasoning in Large Language Models

Liu, Shuaitong, Li, Renjue, Yu, Lijia, Zhang, Lijun, Liu, Zhiming, Jin, Gaojie

arXiv.org Artificial Intelligence

Recent advances in Chain-of-Thought (CoT) prompting have substantially improved the reasoning capabilities of large language models (LLMs), but have also introduced their computational efficiency as a new attack surface. In this paper, we propose BadThink, the first backdoor attack designed to deliberately induce "overthinking" behavior in CoT-enabled LLMs while ensuring stealth. When activated by carefully crafted trigger prompts, BadThink manipulates the model to generate inflated reasoning traces - producing unnecessarily redundant thought processes while preserving the consistency of final outputs. This subtle attack vector creates a covert form of performance degradation that significantly increases computational costs and inference time while remaining difficult to detect through conventional output evaluation methods. We implement this attack through a sophisticated poisoning-based fine-tuning strategy, employing a novel LLM-based iterative optimization process to embed the behavior by generating highly naturalistic poisoned data. Our experiments on multiple state-of-the-art models and reasoning tasks show that BadThink consistently increases reasoning trace lengths - achieving an over 17x increase on the MATH-500 dataset - while remaining stealthy and robust. This work reveals a critical, previously unexplored vulnerability where reasoning efficiency can be covertly manipulated, demonstrating a new class of sophisticated attacks against CoT-enabled systems.